402 research outputs found

    Approximating the Distribution of the Median and other Robust Estimators on Uncertain Data

    Get PDF
    Robust estimators, like the median of a point set, are important for data analysis in the presence of outliers. We study robust estimators for locationally uncertain points with discrete distributions. That is, each point in a data set has a discrete probability distribution describing its location. The probabilistic nature of uncertain data makes it challenging to compute such estimators, since the true value of the estimator is now described by a distribution rather than a single point. We show how to construct and estimate the distribution of the median of a point set. Building the approximate support of the distribution takes near-linear time, and assigning probability to that support takes quadratic time. We also develop a general approximation technique for distributions of robust estimators with respect to ranges with bounded VC dimension. This includes the geometric median for high dimensions and the Siegel estimator for linear regression.Comment: Full version of a paper to appear at SoCG 201

    Finding a minimum stretch of a function

    Get PDF
    Given a piecewise monotone function f : R ! R and a real value Tmin, we develop an algorithm that finds an interval of length at least Tmin for which the average value of f is minimized. The run-time of the algorithm is linear in the number of monotone pieces of f if certain operations are available in constant time for f. We use this algorithm to solve a basic problem arising in the analysis of trajectories: Finding the most similar subtrajectories of two given trajectories, provided that the duration is at least Tmin. Since the precise solution requires complex operations, we also give a simple (1+")approximation algorithm in which these operations are not needed

    Finding long and similar parts of trajectories

    Full text link
    A natural time-dependent similarity measure for two trajectories is their average distance at corresponding times. We give algorithms for computing the most similar subtrajectories under this measure, assuming the two trajectories are given as two polygonal, possibly self-intersecting lines. When a minimum duration is specified for the subtrajectories, and they must start at exactly corresponding times in the input trajectories, we give a linear-time algorithm for computing the starting time and duration of the most similar subtrajectories. The algorithm is based on a result of independent interest: We present a linear-time algorithm to find, for a piece-wise monotone function, an interval of at least a given length that has minimum average value. When the two subtrajectories can start at different times in the two input trajectories, it appears difficult to give an exact algorithm for the most similar subtrajectories problem, even if the duration of the desired two subtrajectories is fixed to some length. We show that the problem can be solved approximately, and with a performance guarantee. More precisely, we present (1 + e)-approximation algorithms for computing the most similar subtrajectories of two input trajectories for the case where the duration is specified, and also for the case where only a minimum on the duration is specified

    Model-based segmentation and classification of trajectories (Extended abstract)

    Get PDF
    We present efficient algorithms for segmenting and classifying a trajectory based on a parameterized movement model like the Brownian bridge movement model. Segmentation is the problem of subdividing a trajectory into parts such that each art is homogeneous in its movement characteristics. We formalize this using the likelihood of the model parameter. We consider the case where a discrete set of m parameter values is given and present an algorithm to compute an optimal segmentation with respect to an information criterion in O(nm) time for a trajectory with n sampling points. Classification is the problem of assigning trajectories to classes. We present an algorithm for discrete classification given a set of trajectories. Our algorithm computes the optimal classification with respect to an information criterion in O(m^2 + mk(log m + log k)) time for m parameter values and k trajectories, assuming bitonic likelihood functions

    Four Soviets walk the dog, with an application to Alt's conjecture

    Get PDF
    Given two polygonal curves in the plane, there are many ways to define a notion of similarity between them. One measure that is extremely popular is the Fréchet distance. Since it has been proposed by Alt and Godau in 1992, many variants and extensions have been studied. Nonetheless, even more than 20 years later, the original O(n^2 log n) algorithm by Alt and Godau for computing the Fréchet distance remains the state of the art (here n denotes the number of vertices on each curve). This has led Helmut Alt to conjecture that the associated decision problem is 3SUM-hard.In recent work, Agarwal et al. show how to break the quadratic barrier for the discrete version of the Fréchet distance, where one considers sequences of points instead of polygonal curves. Building on their work, we give a randomized algorithm to compute the Fréchet distance between two polygonal curves in time O(n^2 \sqrt log n (log log n)^{3/2}) on a pointer machine and in time O(n^2 (log log n)^2) on a word RAM. Furthermore, we show that there exists an algebraic decision tree for the decision problem of depth O(n^{2¿}), for some ¿ > 0. This provides evidence that the decision problem may not be 3SUM-hard after all and reveals an intriguing new aspect of this well-studied problem

    Fine-Grained Complexity Analysis of Two Classic TSP Variants

    Get PDF
    We analyze two classic variants of the Traveling Salesman Problem using the toolkit of fine-grained complexity. Our first set of results is motivated by the Bitonic TSP problem: given a set of nn points in the plane, compute a shortest tour consisting of two monotone chains. It is a classic dynamic-programming exercise to solve this problem in O(n2)O(n^2) time. While the near-quadratic dependency of similar dynamic programs for Longest Common Subsequence and Discrete Frechet Distance has recently been proven to be essentially optimal under the Strong Exponential Time Hypothesis, we show that bitonic tours can be found in subquadratic time. More precisely, we present an algorithm that solves bitonic TSP in O(nlog2n)O(n \log^2 n) time and its bottleneck version in O(nlog3n)O(n \log^3 n) time. Our second set of results concerns the popular kk-OPT heuristic for TSP in the graph setting. More precisely, we study the kk-OPT decision problem, which asks whether a given tour can be improved by a kk-OPT move that replaces kk edges in the tour by kk new edges. A simple algorithm solves kk-OPT in O(nk)O(n^k) time for fixed kk. For 2-OPT, this is easily seen to be optimal. For k=3k=3 we prove that an algorithm with a runtime of the form O~(n3ϵ)\tilde{O}(n^{3-\epsilon}) exists if and only if All-Pairs Shortest Paths in weighted digraphs has such an algorithm. The results for k=2,3k=2,3 may suggest that the actual time complexity of kk-OPT is Θ(nk)\Theta(n^k). We show that this is not the case, by presenting an algorithm that finds the best kk-move in O(n2k/3+1)O(n^{\lfloor 2k/3 \rfloor + 1}) time for fixed k3k \geq 3. This implies that 4-OPT can be solved in O(n3)O(n^3) time, matching the best-known algorithm for 3-OPT. Finally, we show how to beat the quadratic barrier for k=2k=2 in two important settings, namely for points in the plane and when we want to solve 2-OPT repeatedly.Comment: Extended abstract appears in the Proceedings of the 43rd International Colloquium on Automata, Languages, and Programming (ICALP 2016

    Computing the Fréchet distance with shortcuts is NP-hard

    Get PDF
    We study the shortcut Fréchet distance, a natural variant of the Fréchet distance that allows us to take shortcuts from and to any point along one of the curves. We show that, surprisingly, the problem of computing the shortcut Fréchet distance exactly is NP-hard. Furthermore, we give a 3-approximation algorithm for the decision version of the problem

    Supporting Intelligent and Trustworthy Maritime Path Planning Decisions

    Get PDF
    The risk of maritime collisions and groundings has dramatically increased in the past five years despite technological advancements such as GPS-based navigation tools and electronic charts which may add to, instead of reduce, workload. We propose that an automated path planning tool for littoral navigation can reduce workload and improve overall system efficiency, particularly under time pressure. To this end, a Maritime Automated Path Planner (MAPP) was developed, incorporating information requirements developed from a cognitive task analysis, with special emphasis on designing for trust. Human-in-the-loop experimental results showed that MAPP was successful in reducing the time required to generate an optimized path, as well as reducing path lengths. The results also showed that while users gave the tool high acceptance ratings, they rated the MAPP as average for trust, which we propose is the appropriate level of trust for such a system.This work was sponsored by Rite Solutions Inc., Assett Inc., Mikel Inc., and the Office of Naval Research. We would also like to thank Northeast Maritime Institute, the MIT NROTC detachment, the crew of the USS New Hampshire, and the anonymous reviewers whose comments significantly improved the paper

    Computing the Fréchet Distance with a Retractable Leash

    Get PDF
    All known algorithms for the Fréchet distance between curves proceed in two steps: first, they construct an efficient oracle for the decision version; second, they use this oracle to find the optimum from a finite set of critical values. We present a novel approach that avoids the detour through the decision version. This gives the first quadratic time algorithm for the Fréchet distance between polygonal curves in (Formula presented.) under polyhedral distance functions (e.g., (Formula presented.) and (Formula presented.)). We also get a (Formula presented.)-approximation of the Fréchet distance under the Euclidean metric, in quadratic time for any fixed (Formula presented.). For the exact Euclidean case, our framework currently yields an algorithm with running time (Formula presented.). However, we conjecture that it may eventually lead to a faster exact algorithm

    EPR-based ghost imaging using a single-photon-sensitive camera

    Get PDF
    Correlated photon imaging, popularly known as ghost imaging, is a technique whereby an image is formed from light that has never interacted with the object. In ghost imaging experiments, two correlated light fields are produced. One of these fields illuminates the object, and the other field is measured by a spatially resolving detector. In the quantum regime, these correlated light fields are produced by entangled photons created by spontaneous parametric down-conversion. To date, all correlated photon ghost imaging experiments have scanned a single-pixel detector through the field of view to obtain spatial information. However, scanning leads to poor sampling efficiency, which scales inversely with the number of pixels, N, in the image. In this work, we overcome this limitation by using a time-gated camera to record the single-photon events across the full scene. We obtain high-contrast images, 90%, in either the image plane or the far field of the photon pair source, taking advantage of the Einstein–Podolsky–Rosen-like correlations in position and momentum of the photon pairs. Our images contain a large number of modes, >500, creating opportunities in low-light-level imaging and in quantum information processing
    corecore